from transformers import AutoModelForCausalLM, AutoTokenizer
tokenizer = AutoTokenizer.from_pretrained('/root/howard/replit-code-v1-3b', trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained('/root/howard/replit-code-v1-3b', trust_remote_code=True)
x = tokenizer.encode('# 写一个快速排序', return_tensors='pt')
x = tokenizer.encode('func (h *wsHandler) handle(w http.ResponseWriter, r *http.Request) {\n applog.Debugw("handle ws request", "url", r.URL.String(), "header", r.Header)\n wsID := r.URL.Query().Get("WorkspaceID")\n # 根据 wsID 查询 url,反向代理', return_tensors='pt')
x = tokenizer.encode('func fibonacci(n int) {', return_tensors='pt')
x = tokenizer.encode('# 写一个 http server ', return_tensors='pt')
y = model.generate(x, max_length=200, do_sample=True, top_p=0.95, top_k=4, temperature=0.8, num_return_sequences=1, eos_token_id=tokenizer.eos_token_id)
# decoding, clean_up_tokenization_spaces=False to ensure syntactical correctness
print(tokenizer.decode(y[0], skip_special_tokens=True, clean_up_tokenization_spaces=False))